# Long context processing
Voxtral Mini 3B 2507 Transformers
Apache-2.0
Voxtral Mini is an enhanced version based on Ministral 3B, with advanced audio input capabilities and excellent performance in speech transcription, translation, and audio understanding.
Audio-to-Text
Transformers Supports Multiple Languages

V
MohamedRashad
416
2
Lucy 128k
Apache-2.0
Lucy is a 1.7-billion parameter model focused on intelligent web search and lightweight browsing. It is built on Qwen3-1.7B and optimized to run efficiently on mobile devices.
Large Language Model
Transformers English

L
Menlo
109
43
Qwen3 4b Tcomanr Merge
A language model merged from multiple Qwen 3 fine-tuned models, integrating code and mathematical abilities to provide powerful language processing capabilities.
Large Language Model
Transformers

Q
ertghiu256
267
1
EXAONE 4.0 1.2B
Other
EXAONE 4.0 is a multilingual large language model that integrates non-inference mode and inference mode, supporting English, Korean, and Spanish, and has the ability to invoke agent tools.
Large Language Model
Transformers Supports Multiple Languages

E
LGAI-EXAONE
189
50
EXAONE 4.0 32B GGUF
Other
The EXAONE 4.0 model integrates non-inference mode and inference mode, combining the excellent usability of EXAONE 3.5 with the advanced inference capabilities of EXAONE Deep. It supports three languages: English, Korean, and Spanish, providing stronger support for intelligent applications.
Large Language Model
Transformers Supports Multiple Languages

E
LGAI-EXAONE
305
17
EXAONE 4.0 32B
Other
EXAONE 4.0 is a multi-functional large language model that integrates non-reasoning and reasoning modes, supports English, Korean, and Spanish, and has the ability to use agent tools.
Large Language Model
Transformers Supports Multiple Languages

E
LGAI-EXAONE
230
145
A.X 3.1 Light
Apache-2.0
A.X 3.1 Light is a lightweight large language model (LLM) optimized for Korean understanding and enterprise deployment, fully independently developed by SKT.
Large Language Model
Transformers Supports Multiple Languages

A
skt
411
26
Smollm3 3B
Apache-2.0
SmolLM3 is a language model with 3 billion parameters, aiming to break the performance boundary of small models. It supports 6 languages, has advanced reasoning ability and long context processing ability, and is a fully open-source model.
Large Language Model
Transformers Supports Multiple Languages

S
HuggingFaceTB
1,138
99
A.X 4.0 Light Gguf
Apache-2.0
A.X 4.0 Light is a lightweight large language model developed by SKT AI Model Lab, built on Qwen2.5 and optimized for Korean understanding and enterprise deployment.
Large Language Model
Transformers Supports Multiple Languages

A
mykor
535
2
Longwriter Zero 32B I1 GGUF
Apache-2.0
The LongWriter-Zero-32B quantized model is based on the THU-KEG/LongWriter-Zero-32B base model, supports both Chinese and English, and is suitable for long context scenarios such as reinforcement learning and writing.
Large Language Model
Transformers Supports Multiple Languages

L
mradermacher
135
1
Longwriter Zero 32B GGUF
Apache-2.0
The LongWriter-Zero-32B Quantized Model is a multilingual model that undergoes static quantization based on the original model. It is suitable for long context scenarios such as reinforcement learning and writing.
Large Language Model
Transformers Supports Multiple Languages

L
mradermacher
204
1
Smollm3 3B Base
Apache-2.0
SmolLM3 is an open-source language model with 3 billion parameters, supporting 6 languages and possessing powerful inference and long context processing capabilities.
Large Language Model
Transformers Supports Multiple Languages

S
HuggingFaceTB
1,208
41
Jan Nano Gguf
Apache-2.0
Jan Nano is a small language model fine - tuned based on the Qwen3 architecture, designed for local and embedded environments, combining efficiency and long - context processing capabilities.
Large Language Model
J
Menlo
2,703
38
Deepseek R1 0528 GPTQ Int4 Int8Mix Compact
MIT
The GPTQ quantized version of the DeepSeek-R1-0528 model, using a quantization scheme of Int4 + selective Int8, which reduces the file size while ensuring the generation quality.
Large Language Model
Transformers

D
QuantTrio
258
1
Zlatorog 12B Instruct Beta
Zlatorog-12B-Instruct-Beta is a large language model fine-tuned based on Mistral Nemo 12B, designed specifically for Slovenian language-related research, supporting instruction following, reasoning, and tool-enhanced generation tasks.
Large Language Model
Transformers Supports Multiple Languages

Z
zID4si
123
1
Phi 4 Reasoning Unsloth Bnb 4bit
MIT
Phi-4-reasoning is an advanced reasoning model developed by Microsoft, fine-tuned based on Phi-4, focusing on improving reasoning abilities in fields such as mathematics, science, and coding.
Large Language Model
Transformers Supports Multiple Languages

P
unsloth
1,969
2
Qwen3
Apache-2.0
Qwen3 is a series of advanced language models developed by the Qwen Team, excelling in advanced reasoning, dialogue, instruction following, and agent applications.
Large Language Model
Q
cortexso
8,089
2
Bamba 9B V2
Apache-2.0
Bamba-9B-v2 is a decoder-only language model built on the Mamba-2 architecture, focusing on text generation tasks and outperforming Llama 3.1 8B in performance.
Large Language Model
Transformers

B
ibm-ai-platform
3,634
15
Llama Krikri 8B Instruct GGUF
A Greek instruction-tuned large language model based on Llama-3.1-8B, enhancing Greek language capabilities and supporting multilingual tasks
Large Language Model
Transformers

L
ilsp
257
11
Llama Krikri 8B Base
Llama-Krikri-8B-Base is a large Greek foundational language model built upon Llama-3.1-8B, extending Greek language capabilities through continued pretraining while maintaining English proficiency.
Large Language Model
Transformers Supports Multiple Languages

L
ilsp
104
12
Qwen2.5 14B DeepSeek R1 1M
Apache-2.0
A multi-purpose large language model that combines the advantages of reasoning models and long context models
Large Language Model
Transformers

Q
mkurman
2,195
52
Deepseek R1 Bf16
MIT
DeepSeek-R1 is the first-generation inference model, which performs excellently in mathematics, code, and reasoning tasks, and its performance is comparable to that of OpenAI-o1.
Large Language Model
Transformers

D
opensourcerelease
1,486
16
Lucie 7B
Apache-2.0
Lucie-7B is a multilingual causal language model with 7 billion parameters jointly built by LINAGORA and OpenLLM-France. It is based on the Llama-3 architecture and pre-trained on 3 trillion tokens of multilingual data.
Large Language Model
TensorBoard Supports Multiple Languages

L
OpenLLM-France
1,262
20
Yi 1.5 9B Chat 16K Abliterated
Apache-2.0
Based on the ablated version of the large language model, with better performance than the original model, supporting Chinese, English, and long context processing.
Large Language Model
Transformers Supports Multiple Languages

Y
byroneverson
10.52k
5
AI21 Jamba Large 1.5
Other
AI21 Jamba 1.5 is a series of advanced foundation models with powerful long context processing capabilities and efficient inference speed, suitable for various business scenarios.
Large Language Model
Safetensors Supports Multiple Languages
A
ai21labs
2,642
216
Jais Adapted 7b
Apache-2.0
The Jais series are specialized bilingual large language models for Arabic processing, adapted for Arabic through pre-training based on the Llama-2 architecture
Large Language Model Supports Multiple Languages
J
inceptionai
1,726
6
Jais Family 590m
Apache-2.0
The Jais series is a bilingual large language model specifically optimized for Arabic while maintaining strong English capabilities. This 590M parameter version is the smallest model in the series, suitable for lightweight applications.
Large Language Model Supports Multiple Languages
J
inceptionai
86
7
Jais Adapted 70b
Apache-2.0
The Jais series is a set of English-Arabic bilingual large language models, optimized for Arabic while maintaining strong English capabilities, with models ranging from 590 million to 70 billion parameters.
Large Language Model Supports Multiple Languages
J
inceptionai
2,579
21
Jais Adapted 70b Chat
Apache-2.0
The Jais series is a large English-Arabic bilingual language model optimized specifically for Arabic. It undergoes adaptive pre-training based on the Llama-2 architecture and has powerful bilingual processing capabilities.
Large Language Model Supports Multiple Languages
J
inceptionai
5,466
9
Meltemi 7B Instruct V1.5
Apache-2.0
Meltemi 7B Instruct v1.5 is a Greek instruction fine-tuned large language model improved based on Mistral 7B, focusing on Greek natural language processing tasks.
Large Language Model
Transformers

M
ilsp
1,237
21
Meltemi 7B V1.5
Apache-2.0
Meltemi is a large-scale Greek basic language model trained by the Athens Center for Research and Innovation, built on Mistral 7B and focused on Greek language processing capabilities.
Large Language Model
Transformers Supports Multiple Languages

M
ilsp
106
15
Aya 23 35B
Aya 23 is an instruction-fine-tuned open-weight research version model with highly advanced multilingual capabilities, supporting 23 languages.
Large Language Model
Transformers Supports Multiple Languages

A
CohereLabs
3,721
282
Evo 1 8k Base
Apache-2.0
Evo is a biological foundation model capable of long context modeling and design. It uses the StripedHyena architecture and can model sequences at single nucleotide and byte-level resolution.
Molecular Model
Transformers

E
togethercomputer
31.09k
9
Yi 34B 200K
Apache-2.0
The Yi series of models are next-generation open-source large language models trained from scratch by 01.AI. They support bilingual (Chinese and English) and perform excellently in language understanding, common-sense reasoning, reading comprehension, etc.
Large Language Model
Transformers

Y
01-ai
12.63k
317
Extended Mind Mpt 7b
An extended mind Transformer model adjusted based on the Mosaic ML architecture, supporting external memory library retrieval and attention functions
Large Language Model
Transformers

E
normalcomputing
111
29
Featured Recommended AI Models